124 research outputs found

    EviPrompt: A Training-Free Evidential Prompt Generation Method for Segment Anything Model in Medical Images

    Full text link
    Medical image segmentation has immense clinical applicability but remains a challenge despite advancements in deep learning. The Segment Anything Model (SAM) exhibits potential in this field, yet the requirement for expertise intervention and the domain gap between natural and medical images poses significant obstacles. This paper introduces a novel training-free evidential prompt generation method named EviPrompt to overcome these issues. The proposed method, built on the inherent similarities within medical images, requires only a single reference image-annotation pair, making it a training-free solution that significantly reduces the need for extensive labeling and computational resources. First, to automatically generate prompts for SAM in medical images, we introduce an evidential method based on uncertainty estimation without the interaction of clinical experts. Then, we incorporate the human prior into the prompts, which is vital for alleviating the domain gap between natural and medical images and enhancing the applicability and usefulness of SAM in medical scenarios. EviPrompt represents an efficient and robust approach to medical image segmentation, with evaluations across a broad range of tasks and modalities confirming its efficacy

    Learnable Descent Algorithm for Nonsmooth Nonconvex Image Reconstruction

    Full text link
    We propose a general learning based framework for solving nonsmooth and nonconvex image reconstruction problems. We model the regularization function as the composition of the l2,1l_{2,1} norm and a smooth but nonconvex feature mapping parametrized as a deep convolutional neural network. We develop a provably convergent descent-type algorithm to solve the nonsmooth nonconvex minimization problem by leveraging the Nesterov's smoothing technique and the idea of residual learning, and learn the network parameters such that the outputs of the algorithm match the references in training data. Our method is versatile as one can employ various modern network structures into the regularization, and the resulting network inherits the guaranteed convergence of the algorithm. We also show that the proposed network is parameter-efficient and its performance compares favorably to the state-of-the-art methods in a variety of image reconstruction problems in practice

    A Learnable Variational Model for Joint Multimodal MRI Reconstruction and Synthesis

    Full text link
    Generating multi-contrasts/modal MRI of the same anatomy enriches diagnostic information but is limited in practice due to excessive data acquisition time. In this paper, we propose a novel deep-learning model for joint reconstruction and synthesis of multi-modal MRI using incomplete k-space data of several source modalities as inputs. The output of our model includes reconstructed images of the source modalities and high-quality image synthesized in the target modality. Our proposed model is formulated as a variational problem that leverages several learnable modality-specific feature extractors and a multimodal synthesis module. We propose a learnable optimization algorithm to solve this model, which induces a multi-phase network whose parameters can be trained using multi-modal MRI data. Moreover, a bilevel-optimization framework is employed for robust parameter training. We demonstrate the effectiveness of our approach using extensive numerical experiments.Comment: 12 page

    Uncertainty-Induced Transferability Representation for Source-Free Unsupervised Domain Adaptation

    Full text link
    Source-free unsupervised domain adaptation (SFUDA) aims to learn a target domain model using unlabeled target data and the knowledge of a well-trained source domain model. Most previous SFUDA works focus on inferring semantics of target data based on the source knowledge. Without measuring the transferability of the source knowledge, these methods insufficiently exploit the source knowledge, and fail to identify the reliability of the inferred target semantics. However, existing transferability measurements require either source data or target labels, which are infeasible in SFUDA. To this end, firstly, we propose a novel Uncertainty-induced Transferability Representation (UTR), which leverages uncertainty as the tool to analyse the channel-wise transferability of the source encoder in the absence of the source data and target labels. The domain-level UTR unravels how transferable the encoder channels are to the target domain and the instance-level UTR characterizes the reliability of the inferred target semantics. Secondly, based on the UTR, we propose a novel Calibrated Adaption Framework (CAF) for SFUDA, including i)the source knowledge calibration module that guides the target model to learn the transferable source knowledge and discard the non-transferable one, and ii)the target semantics calibration module that calibrates the unreliable semantics. With the help of the calibrated source knowledge and the target semantics, the model adapts to the target domain safely and ultimately better. We verified the effectiveness of our method using experimental results and demonstrated that the proposed method achieves state-of-the-art performances on the three SFUDA benchmarks. Code is available at https://github.com/SPIresearch/UTR

    Fuel Consumption Evaluation of Connected Automated Vehicles Under Rear-End Collisions

    Get PDF
    Connected automated vehicles (CAV) can increase traffic efficiency, which is considered a critical factor in saving energy and reducing emissions in traffic congestion. In this paper, systematic traffic simulations are conducted for three car-following modes, including intelligent driver model (IDM), adaptive cruise control (ACC), and cooperative ACC (CACC), in congestions caused by rear-end collisions. From the perspectives of lane density, vehicle trajectory and vehicle speed, the fuel consumption of vehicles under the three car-following modes are compared and analysed, respectively. Based on the vehicle driving and accident environment parameters, an XGBoost algorithm-based fuel consumption prediction framework is proposed for traffic congestions caused by rear-end collisions. The results show that compared with IDM and ACC modes, the vehicles in CACC car-following mode have the ideal performance in terms of total fuel consumption; besides, the traffic flow in CACC mode is more stable, and the speed fluctuation is relatively tiny in different accident impact regions, which meets the driving desires of drivers

    Fuzzing Deep Learning Compilers with HirGen

    Full text link
    Deep Learning (DL) compilers are widely adopted to optimize advanced DL models for efficient deployment on diverse hardware. Their quality has profound effect on the quality of compiled DL models. A recent bug study shows that the optimization of high-level intermediate representation (IR) is the most error-prone compilation stage. Bugs in this stage are accountable for 44.92% of the whole collected ones. However, existing testing techniques do not consider high-level optimization related features (e.g. high-level IR), and are therefore weak in exposing bugs at this stage. To bridge this gap, we propose HirGen, an automated testing technique that aims to effectively expose coding mistakes in the optimization of high-level IR. The design of HirGen includes 1) three coverage criteria to generate diverse and valid computational graphs; 2) full use of high-level IRs language features to generate diverse IRs; 3) three test oracles inspired from both differential testing and metamorphic testing. HirGen has successfully detected 21 bugs that occur at TVM, with 17 bugs confirmed and 12 fixed. Further, we construct four baselines using the state-of-the-art DL compiler fuzzers that can cover the high-level optimization stage. Our experiment results show that HirGen can detect 10 crashes and inconsistencies that cannot be detected by the baselines in 48 hours. We further validate the usefulness of our proposed coverage criteria and test oracles in evaluation

    DopNet:A Deep Convolutional Neural Network to Recognize Armed and Unarmed Human Targets

    Get PDF
    The work presented in this paper aims to distinguish between armed or unarmed personnel using multi-static radar data and advanced Doppler processing. We propose two modified Deep Convolutional Neural Networks (DCNN) termed SCDopNet and MC-DopNet for mono-static and multi-static micro- Doppler signature (μ-DS) classification. Differentiating armed and unarmed walking personnel is challenging due to the effect of aspect angle and channel diversity in real-world scenarios. In addition, DCNN easily overfits the relatively small-scale μ-DS dataset. To address these problems, the work carried out in this paper makes three key contributions: first, two effective schemes including data augmentation operation and a regularization term are proposed to train SC-DopNet from scratch. Next, a factor analysis of the SC-DopNet are conducted based on various operating parameters in both the processing and radar operations. Thirdly, to solve the problem of aspect angle diversity for μ-DS classification, we design MC-DopNet for multi-static μ- DS which is embedded with two new fusion schemes termed as Greedy Importance Reweighting (GIR) and `21-Norm. These two schemes are based on two different strategies and have been evaluated experimentally: GIR uses a “win by sacrificing worst case” whilst `21-Norm adopts a “win by sacrificing best case” approach. The SC-DopNet outperforms the non-deep methods by 12.5% in average and the proposed MC-DopNet with two fusion methods outperforms the conventional binary voting by 1.2% in average. Note that we also argue and discuss how to utilize the statistics of SC-DopNet results to infer the selection of fusion strategies for MC-DopNet under different experimental scenarios
    corecore